12 - Optimization-based control for large-scale and complex systems: When and why does it work? [ID:58035]
50 von 706 angezeigt

Thank you, Enrique, for this kind introduction.

Yeah, that's precisely the summary of what I'm going to talk about.

I would like to talk about challenging optimal control problems from practice, from academia,

and how one can solve them even when they are very large.

And I would like to spend the first half of my talk on modern predictive control and the

second half of my talk on reinforcement learning.

This is material that has evolved over the last 10, 15 years, so there are a lot of coworkers.

I would like to point out two coworkers for this use case, which is a real industrial

use case, which is by Matthias Höger, who is a former student of mine, and Kilian Link,

who are both at Siemens Energy here in Erlang and just 500 meters away.

Yeah, so let me start with a brief introduction.

What is this all about?

This is about optimal control.

So we need some control system, and let me start right with a formula.

So this is a control system here in continuous time, where we have a differential equation.

This is a control system in discrete time, where G defines the next state at a discrete

time instant.

And for this talk, it doesn't really matter.

It will apply to both, more or less.

I will, of course, not do everything for both models, but I will switch from one to the

other occasionally, whatever is more convenient.

Okay, so X is the state.

This is what we want to control.

U is the control input.

That's the variable time depending that we are allowed to change according to our needs.

And many control problems, particularly if they last longer than just a few seconds or

whatever the dynamic scale is, will require optimal control in feedback form.

What's a feedback control?

A feedback control means that we do not pre-specify this control U of t as a variable or as a

function of t for a long time into the future, but actually would like to have our control

as a map F that takes the state or a measurement of the state and produces the control value

from that.

And the main feedback is completely clear.

We look at what the system is doing, and this feedback from the system tells us how to control

it in the new.

Yeah, the reason for that is that with the feedback, we are much more able to correct

deviations and deviations are everywhere.

We have modeling errors.

We have perturbations that we have not foreseen.

And just to show you a little quick example, it's this one.

This is a pendulum example that a student of mine did a while ago.

So it's a swing up of the pendulum.

This, of course, you could do with open loop, but then the feedback is needed because the

student is going to tease the pendulum and the tensor realizes that the pendulum is pushed

and reacts to that by swinging it up again if necessary and making correct.

That's the mechanism of feedback that is what provides the robustness that we want to have

in most practical applications.

Okay.

Now, I would like to talk about control problems on infinite horizon today.

There are many interesting control problems that are on finite horizons, but I would like

Teil einer Videoserie :

Presenters

Prof. Dr. Lars Grüne Prof. Dr. Lars Grüne

Zugänglich über

Offener Zugang

Dauer

00:55:05 Min

Aufnahmedatum

2025-06-03

Hochgeladen am

2025-06-04 09:39:09

Sprache

en-US

Tags

FAU FAU MoD FAU MoD Lecture FAU MoD Lecture Series
Einbetten
Wordpress FAU Plugin
iFrame
Teilen